124 research outputs found
EviPrompt: A Training-Free Evidential Prompt Generation Method for Segment Anything Model in Medical Images
Medical image segmentation has immense clinical applicability but remains a
challenge despite advancements in deep learning. The Segment Anything Model
(SAM) exhibits potential in this field, yet the requirement for expertise
intervention and the domain gap between natural and medical images poses
significant obstacles. This paper introduces a novel training-free evidential
prompt generation method named EviPrompt to overcome these issues. The proposed
method, built on the inherent similarities within medical images, requires only
a single reference image-annotation pair, making it a training-free solution
that significantly reduces the need for extensive labeling and computational
resources. First, to automatically generate prompts for SAM in medical images,
we introduce an evidential method based on uncertainty estimation without the
interaction of clinical experts. Then, we incorporate the human prior into the
prompts, which is vital for alleviating the domain gap between natural and
medical images and enhancing the applicability and usefulness of SAM in medical
scenarios. EviPrompt represents an efficient and robust approach to medical
image segmentation, with evaluations across a broad range of tasks and
modalities confirming its efficacy
Learnable Descent Algorithm for Nonsmooth Nonconvex Image Reconstruction
We propose a general learning based framework for solving nonsmooth and
nonconvex image reconstruction problems. We model the regularization function
as the composition of the norm and a smooth but nonconvex feature
mapping parametrized as a deep convolutional neural network. We develop a
provably convergent descent-type algorithm to solve the nonsmooth nonconvex
minimization problem by leveraging the Nesterov's smoothing technique and the
idea of residual learning, and learn the network parameters such that the
outputs of the algorithm match the references in training data. Our method is
versatile as one can employ various modern network structures into the
regularization, and the resulting network inherits the guaranteed convergence
of the algorithm. We also show that the proposed network is parameter-efficient
and its performance compares favorably to the state-of-the-art methods in a
variety of image reconstruction problems in practice
A Learnable Variational Model for Joint Multimodal MRI Reconstruction and Synthesis
Generating multi-contrasts/modal MRI of the same anatomy enriches diagnostic
information but is limited in practice due to excessive data acquisition time.
In this paper, we propose a novel deep-learning model for joint reconstruction
and synthesis of multi-modal MRI using incomplete k-space data of several
source modalities as inputs. The output of our model includes reconstructed
images of the source modalities and high-quality image synthesized in the
target modality. Our proposed model is formulated as a variational problem that
leverages several learnable modality-specific feature extractors and a
multimodal synthesis module. We propose a learnable optimization algorithm to
solve this model, which induces a multi-phase network whose parameters can be
trained using multi-modal MRI data. Moreover, a bilevel-optimization framework
is employed for robust parameter training. We demonstrate the effectiveness of
our approach using extensive numerical experiments.Comment: 12 page
Uncertainty-Induced Transferability Representation for Source-Free Unsupervised Domain Adaptation
Source-free unsupervised domain adaptation (SFUDA) aims to learn a target
domain model using unlabeled target data and the knowledge of a well-trained
source domain model. Most previous SFUDA works focus on inferring semantics of
target data based on the source knowledge. Without measuring the
transferability of the source knowledge, these methods insufficiently exploit
the source knowledge, and fail to identify the reliability of the inferred
target semantics. However, existing transferability measurements require either
source data or target labels, which are infeasible in SFUDA. To this end,
firstly, we propose a novel Uncertainty-induced Transferability Representation
(UTR), which leverages uncertainty as the tool to analyse the channel-wise
transferability of the source encoder in the absence of the source data and
target labels. The domain-level UTR unravels how transferable the encoder
channels are to the target domain and the instance-level UTR characterizes the
reliability of the inferred target semantics. Secondly, based on the UTR, we
propose a novel Calibrated Adaption Framework (CAF) for SFUDA, including i)the
source knowledge calibration module that guides the target model to learn the
transferable source knowledge and discard the non-transferable one, and ii)the
target semantics calibration module that calibrates the unreliable semantics.
With the help of the calibrated source knowledge and the target semantics, the
model adapts to the target domain safely and ultimately better. We verified the
effectiveness of our method using experimental results and demonstrated that
the proposed method achieves state-of-the-art performances on the three SFUDA
benchmarks. Code is available at https://github.com/SPIresearch/UTR
Fuel Consumption Evaluation of Connected Automated Vehicles Under Rear-End Collisions
Connected automated vehicles (CAV) can increase traffic efficiency, which is considered a critical factor in saving energy and reducing emissions in traffic congestion. In this paper, systematic traffic simulations are conducted for three car-following modes, including intelligent driver model (IDM), adaptive cruise control (ACC), and cooperative ACC (CACC), in congestions caused by rear-end collisions. From the perspectives of lane density, vehicle trajectory and vehicle speed, the fuel consumption of vehicles under the three car-following modes are compared and analysed, respectively. Based on the vehicle driving and accident environment parameters, an XGBoost algorithm-based fuel consumption prediction framework is proposed for traffic congestions caused by rear-end collisions. The results show that compared with IDM and ACC modes, the vehicles in CACC car-following mode have the ideal performance in terms of total fuel consumption; besides, the traffic flow in CACC mode is more stable, and the speed fluctuation is relatively tiny in different accident impact regions, which meets the driving desires of drivers
Fuzzing Deep Learning Compilers with HirGen
Deep Learning (DL) compilers are widely adopted to optimize advanced DL
models for efficient deployment on diverse hardware. Their quality has profound
effect on the quality of compiled DL models. A recent bug study shows that the
optimization of high-level intermediate representation (IR) is the most
error-prone compilation stage. Bugs in this stage are accountable for 44.92% of
the whole collected ones. However, existing testing techniques do not consider
high-level optimization related features (e.g. high-level IR), and are
therefore weak in exposing bugs at this stage. To bridge this gap, we propose
HirGen, an automated testing technique that aims to effectively expose coding
mistakes in the optimization of high-level IR. The design of HirGen includes 1)
three coverage criteria to generate diverse and valid computational graphs; 2)
full use of high-level IRs language features to generate diverse IRs; 3) three
test oracles inspired from both differential testing and metamorphic testing.
HirGen has successfully detected 21 bugs that occur at TVM, with 17 bugs
confirmed and 12 fixed. Further, we construct four baselines using the
state-of-the-art DL compiler fuzzers that can cover the high-level optimization
stage. Our experiment results show that HirGen can detect 10 crashes and
inconsistencies that cannot be detected by the baselines in 48 hours. We
further validate the usefulness of our proposed coverage criteria and test
oracles in evaluation
DopNet:A Deep Convolutional Neural Network to Recognize Armed and Unarmed Human Targets
The work presented in this paper aims to distinguish
between armed or unarmed personnel using multi-static radar
data and advanced Doppler processing. We propose two modified
Deep Convolutional Neural Networks (DCNN) termed SCDopNet
and MC-DopNet for mono-static and multi-static micro-
Doppler signature (μ-DS) classification. Differentiating armed
and unarmed walking personnel is challenging due to the effect
of aspect angle and channel diversity in real-world scenarios.
In addition, DCNN easily overfits the relatively small-scale μ-DS
dataset. To address these problems, the work carried out in this
paper makes three key contributions: first, two effective schemes
including data augmentation operation and a regularization
term are proposed to train SC-DopNet from scratch. Next,
a factor analysis of the SC-DopNet are conducted based on
various operating parameters in both the processing and radar
operations. Thirdly, to solve the problem of aspect angle diversity
for μ-DS classification, we design MC-DopNet for multi-static μ-
DS which is embedded with two new fusion schemes termed
as Greedy Importance Reweighting (GIR) and `21-Norm. These
two schemes are based on two different strategies and have been
evaluated experimentally: GIR uses a “win by sacrificing worst
case” whilst `21-Norm adopts a “win by sacrificing best case”
approach. The SC-DopNet outperforms the non-deep methods
by 12.5% in average and the proposed MC-DopNet with two
fusion methods outperforms the conventional binary voting by
1.2% in average. Note that we also argue and discuss how to
utilize the statistics of SC-DopNet results to infer the selection
of fusion strategies for MC-DopNet under different experimental
scenarios
- …